Training Conditional Random Fields for Maximum Labelwise Accuracy

نویسندگان

  • Samuel S. Gross
  • Olga Russakovsky
  • Chuong B. Do
  • Serafim Batzoglou
چکیده

We consider the problem of training a conditional random field (CRF) to maximize per-label predictive accuracy on a training set, an approach motivated by the principle of empirical risk minimization. We give a gradient-based procedure for minimizing an arbitrarily accurate approximation of the empirical risk under a Hamming loss function. In experiments with both simulated and real data, our optimization procedure gives significantly better testing performance than several current approaches for CRF training, especially in situations of high label noise.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

AUC-maximized Deep Convolutional Neural Fields for Sequence Labeling

Learning from complex data with imbalanced label distribution is a challenging problem, especially when the data/label form structure, such as linearchain or tree-like. The widely-used training methods, such as maximum-likelihood and maximum labelwise accuracy, do not work well on imbalanced structured data. To model the complex relationship between the data and the structured label, we present...

متن کامل

Hidden Conditional Random Fields for ECG Classification

In this paper a novel approach to ECG signal classification is proposed. The approach is based on using hidden conditional random fields (HCRF) to model the ECG signal. Features used in training and testing the HCRF are based on time-frequency analysis of the ECG waveforms. Experimental results show that the HCRF model is promising and gives higher accuracy compared to maximum-likelihood (ML) t...

متن کامل

Closed Form Maximum Likelihood Estimator Of Conditional Random Fields

Training Conditional Random Fields (CRFs) can be very slow for big data. In this paper, we present a new training method for CRFs called Empirical Training which is motivated by the concept of co-occurrence rate. We show that the standard training (unregularized) can have many maximum likelihood estimations (MLEs). Empirical training has a unique closed form MLE which is also a MLE of the stand...

متن کامل

Training Conditional Random Fields Using Virtual Evidence Boosting

While conditional random fields (CRFs) have been applied successfully in a variety of domains, their training remains a challenging task. In this paper, we introduce a novel training method for CRFs, called virtual evidence boosting, which simultaneously performs feature selection and parameter estimation. To achieve this, we extend standard boosting to handle virtual evidence, where an observa...

متن کامل

Shallow Parsing with Conditional Random Fields

Conditional random fields for sequence labeling offer advantages over both generative models like HMMs and classifiers applied at each sequence position. Among sequence labeling tasks in language processing, shallow parsing has received much attention, with the development of standard evaluation datasets and extensive comparison among methods. We show here how to train a conditional random fiel...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2006